Iterative Low-Rank Approximation for CNN Compression

نویسنده

  • Maksym Kholiavchenko
چکیده

Deep convolutional neural networks contain tens of millions of parameters, making them impossible to work efficiently on embedded devices. We propose iterative approach of applying low-rank approximation to compress deep convolutional neural networks. Since classification and object detection are the most favored tasks for embedded devices, we demonstrate the effectiveness of our approach by compressing AlexNet [1], VGG-16 [2], YOLOv2 [3] and Tiny YOLO networks. Our results show the superiority of the proposed method compared to non-repetitive ones. We demonstrate higher compression ratio providing less accuracy loss.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

On Optimal Low-Rank Approximation of Multidimensional Discrete Signals

This brief describes an algorithmic development of the optimal low-rank approximation (LRA) of multidimensional (M-D) signals with M 3. The algorithms developed can be regarded as a dimensional generalization of the singular value decomposition (SVD) which is of fundamental importance for analyzing signals that can be represented in a matrix form. In particular, iterative algorithms for optimal...

متن کامل

Nonnegative Matrix Factorization without Nonnegativity Constraints on the Factors

Abstract. We consider a new kind of low rank matrix approximation problem for nonnegative matrices: given a nonnegative matrix M , approximate it with a low rank product V.H such that V.H is nonnegative, but without nonnegativity constraints on V and H separately. The nonnegativity constraint on V.H is natural when using the Kullback-Leibler divergence as optimality criterion. We propose an ite...

متن کامل

Non-iterative generalized low rank approximation of matrices

As an extension to 2DPCA, Generalized Low Rank Approximation of Matrices (GLRAM) applies two-sided (i.e., the left and right) rather than single-sided (i.e., the left or the right alone) linear projecting transform(s) to each 2D image for compression and feature extraction. Its advantages over 2DPCA include higher compression ratio and superior classification performance etc. However, GLRAM can...

متن کامل

Towards Convolutional Neural Networks Compression via Global Error Reconstruction

In recent years, convolutional neural networks (CNNs) have achieved remarkable success in various applications such as image classification, object detection, object parsing and face alignment. Such CNN models are extremely powerful to deal with massive amounts of training data by using millions and billions of parameters. However, these models are typically deficient due to the heavy cost in m...

متن کامل

Lossy Color Image Compression Based on Singular Value Decomposition and GNU GZIP

In matrix algebra, the Singular value decomposition (SVD) is an factorization of complex matrix that has been applied to principal component analysis, canonical correlation in statistics, the determination of the low rank approximation of matrices. In this paper, using the SVD and the theory of low rank approximation of a matrix, we offer a new scheme for color image compression based on singul...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2018